skip to main content


Search for: All records

Creators/Authors contains: "Larsen, Michael L."

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. A realistic approach for gathering high-resolution observations of the rainfall rate, R, in the vertical plane is to use data from vertically pointing Doppler radars. After accounting for the vertical air velocity and attenuation, it is possible to determine the fine, spatially resolved drop size spectra and to calculate R for further statistical analyses. The first such results in a vertical plane are reported here. Specifically, we present results using MRR-Pro Doppler radar observations at resolutions of ten meters in height over the lowest 1.28 km, as well as ten seconds in time, over four sets of observations using two different radars at different locations. Both the correlation functions and power spectra are useful for translating observations and numerical model outputs of R from one scale down to other scales that may be more appropriate for particular applications, such as flood warnings and soil erosion, for example. However, it was found in all cases that, while locally applicable radial power spectra could be calculated, because of statistical heterogeneity most of the power spectra lost all generality, and proper correlation functions could not be computed in general except for one 17-min interval. Nevertheless, these results are still useful since they can be combined to develop catalogs of power spectra over different meteorological conditions and in different climatological settings and locations. Furthermore, even with the limitations of these data, this approach is being used to gain a deeper understanding of rainfall to be reported in a forthcoming paper. 
    more » « less
  2. It is important to understand the statistical–physical structure of the rain in the vertical so that observations aloft can be translated meaningfully into what will occur at the surface. In order to achieve this understanding, it is necessary to gather high temporal and spatial resolution observations of rain in the vertical. This can be achieved by translating radar Doppler spectra into drop size distributions. A long-standing difficulty in using such measurements, however, is the problem of vertical air motion, which can shift the Doppler spectra and therefore significantly alter the deduced drop size distributions and integrated variables. In this work, we overcome this difficulty by requiring that the measured radar reflectivity and the calculated rainfall rates satisfy fundamental physical theory. As a consequence, the mean vertical airspeed can be estimated and removed. Application of this new approach is demonstrated using vertically pointing Doppler radar observations in weak convection. It is shown that the new approach produces what appear to be better estimates of the rainfall rates as well as estimates of the temporal and spatial regionally coherent updraft and downdrafts occurring in the precipitation. The technique is readily applicable to other radars, especially those operating at non-attenuating frequencies. 
    more » « less
  3. Abstract

    Droplet-level interactions in clouds are often parameterized by a modified gamma fitted to a “global” droplet size distribution. Do “local” droplet size distributions of relevance to microphysical processes look like these average distributions? This paper describes an algorithm to search and classify characteristic size distributions within a cloud. The approach combines hypothesis testing, specifically, the Kolmogorov–Smirnov (KS) test, and a widely used class of machine learning algorithms for identifying clusters of samples with similar properties: density-based spatial clustering of applications with noise (DBSCAN) is used as the specific example for illustration. The two-sample KS test does not presume any specific distribution, is parameter free, and avoids biases from binning. Importantly, the number of clusters is not an input parameter of the DBSCAN-type algorithms but is independently determined in an unsupervised fashion. As implemented, it works on an abstract space from the KS test results, and hence spatial correlation is not required for a cluster. The method is explored using data obtained from the Holographic Detector for Clouds (HOLODEC) deployed during the Aerosol and Cloud Experiments in the Eastern North Atlantic (ACE-ENA) field campaign. The algorithm identifies evidence of the existence of clusters of nearly identical local size distributions. It is found that cloud segments have as few as one and as many as seven characteristic size distributions. To validate the algorithm’s robustness, it is tested on a synthetic dataset and successfully identifies the predefined distributions at plausible noise levels. The algorithm is general and is expected to be useful in other applications, such as remote sensing of cloud and rain properties.

    Significance Statement

    A typical cloud can have billions of drops spread over tens or hundreds of kilometers in space. Keeping track of the sizes, positions, and interactions of all of these droplets is impractical, and, as such, information about the relative abundance of large and small drops is typically quantified with a “size distribution.” Droplets in a cloud interact locally, however, so this work is motivated by the question of whether the cloud droplet size distribution is different in different parts of a cloud. A new method, based on hypothesis testing and machine learning, determines how many different size distributions are contained in a given cloud. This is important because the size distribution describes processes such as cloud droplet growth and light transmission through clouds.

     
    more » « less
  4. The 2-Dimensional Video Disdrometer (2DVD) is a commonly used tool for exploring rain microphysics and for validating remotely sensed rain retrievals. Recent work has revealed a persistent anomaly in 2DVD data. Early investigations of this anomaly concluded that the resulting errors in rain measurement were modest, but the methods used to flag anomalous data were not optimized, and related considerations associated with the sample sensing area were not fully investigated. Here, we (i) refine the anomaly-detecting algorithm for increased sensitivity and reliability and (ii) develop a related algorithm for refining the estimate of sample sensing area for all detected drops, including those not directly impacted by the anomaly. Using these algorithms, we explore the corrected data to measure any resulting changes to estimates of bulk rainfall statistics from two separate 2DVDs deployed in South Carolina combining for approximately 10 total years of instrumental uptime. Analysis of this data set consisting of over 200 million drops shows that the error induced in estimated total rain accumulations using the manufacturer-reported area is larger than the error due to considerations related to the anomaly. The algorithms presented here imply that approximately 4.2% of detected drops are spurious and the mean reported effective sample area for drops believed to be correctly detected is overestimated by ~8.5%. Simultaneously accounting for all of these effects suggests that the total accumulated rainfall in the data record is approximately 1.1% larger than the raw data record suggests. 
    more » « less
  5. Radiative transfer through clouds can be impacted by variations in particle number size distribution, but also in particle spatial distribution. Due to turbulent mixing and inertial effects, spatial correlations often exist, even on scales reaching the cloud droplet separation distance. The resulting clusters and voids within the droplet field can lead to deviations from exponential extinction. Prior work has numerically investigated these departures from exponential attenuation in absorptive and scattering media; this work takes a step towards determining the feasibility of detecting departures from exponential behavior due to spatial correlation in turbulent clouds generated in a laboratory setting. Large Eddy Simulation (LES) is used to mimic turbulent mixing clouds generated in a laboratory convection cloud chamber. Light propagation through the resulting polydisperse and spatially correlated particle fields is explored via Monte Carlo ray tracing simulations. The key finding is that both mean radiative flux and standard deviation about the mean differ when correlations exist, suggesting that an experiment using a laboratory convection cloud chamber could be designed to investigate non-exponential behavior. Total forward flux is largely unchanged (due to scattering being highly forward-dominant for the size parameters considered), allowing it to be used for conditional sampling based on optical thickness. Direct and diffuse forward flux means are modified by approximately one standard deviation. Standard deviations of diffuse forward and backward fluxes are strongly enhanced, suggesting that fluctuations in the scattered light are a more sensitive metric to consider. The results also suggest the possibility that measurements of radiative transfer could be used to infer the strength and scales of correlations in a turbulent cloud, indicating entrainment and mixing effects. 
    more » « less
  6. Abstract

    Data collected with a holographic instrument [Holographic Detector for Clouds (HOLODEC)] on board the High-Performance Instrumented Airborne Platform for Environmental Research Gulfstream-V (HIAPER GV) aircraft from marine stratocumulus clouds during the Cloud System Evolution in the Trades (CSET) field project are examined for spatial uniformity. During one flight leg at 1190 m altitude, 1816 consecutive holograms were taken, which were approximately 40 m apart with individual hologram dimensions of 1.16 cm × 0.68 cm × 12.0 cm and with droplet concentrations of up to 500 cm−3. Unlike earlier studies, minimally intrusive data processing (e.g., bypassing calculation of number concentrations, binning, and parametric fitting) is used to test for spatial uniformity of clouds on intra- and interhologram spatial scales (a few centimeters and 40 m, respectively). As a means to test this, measured droplet count fluctuations are normalized with the expected standard deviation from theoretical Poisson distributions, which signifies randomness. Despite the absence of trends in the mean concentration, it is found that the null hypothesis of spatial uniformity on both spatial scales can be rejected with compelling statistical confidence. Monte Carlo simulations suggest that weak clustering explains this signature. These findings also hold for size-resolved analysis but with less certainty. Clustering of droplets caused by, for example, entrainment and turbulence, is size dependent and is likely to influence key processes such as droplet growth and thus cloud lifetime.

     
    more » « less